Farkas Lemma
   HOME

TheInfoList



OR:

Farkas' lemma is a solvability theorem for a finite system of linear inequalities in mathematics. It was originally proven by the Hungarian mathematician Gyula Farkas. Farkas' lemma is the key result underpinning the linear programming duality and has played a central role in the development of mathematical optimization (alternatively,
mathematical programming Mathematical optimization (alternatively spelled ''optimisation'') or mathematical programming is the selection of a best element, with regard to some criterion, from some set of available alternatives. It is generally divided into two subfi ...
). It is used amongst other things in the proof of the Karush–Kuhn–Tucker theorem in
nonlinear programming In mathematics, nonlinear programming (NLP) is the process of solving an optimization problem where some of the constraints or the objective function are nonlinear. An optimization problem is one of calculation of the extrema (maxima, minima or s ...
. Remarkably, in the area of the foundations of quantum theory, the lemma also underlies the complete set of Bell inequalities in the form of necessary and sufficient conditions for the existence of a
local hidden-variable theory In the interpretation of quantum mechanics, a local hidden-variable theory is a hidden-variable theory that satisfies the condition of being consistent with local realism. This includes all types of the theory that attempt to account for the proba ...
, given data from any specific set of measurements. Generalizations of the Farkas' lemma are about the solvability theorem for convex inequalities, i.e., infinite system of linear inequalities. Farkas' lemma belongs to a class of statements called "theorems of the alternative": a theorem stating that exactly one of two systems has a solution.


Statement of the lemma

There are a number of slightly different (but equivalent) formulations of the lemma in the literature. The one given here is due to Gale, Kuhn and Tucker (1951). Here, the notation \mathbf \geq 0 means that all components of the vector \mathbf are nonnegative.


Example

Let ''m'', ''n'' = 2, \mathbf = \begin6 & 4 \\3 & 0\end, and \mathbf = \beginb_1 \\b_2\end. The lemma says that exactly one of the following two statements must be true (depending on ''b''1 and ''b''2): # There exist ''x''1 ≥ 0, ''x''2 ≥ 0 such that 6 ''x''1 + 4 ''x''2 = ''b''1 and 3 ''x''1 = ''b''2, or # There exist ''y''1, ''y''2 such that 6 ''y''1 + 3 ''y''2 ≥ 0, 4 ''y''1 ≥ 0, and ''b''1 ''y''1 + ''b''2 ''y''2 < 0. Here is a proof of the lemma in this special case: * If ''b''2 ≥ 0 and ''b''1 − 2''b''2 ≥ 0, then option 1 is true, since the solution of the linear equations is ''x''1 = ''b''2/3 and ''x''2 = (''b''1-2''b''2) / 4. Option 2 is false, since ''b''1 ''y''1 + ''b''2 ''y''2 ≥ ''b''2 (2 ''y''1 + ''y''2) = ''b''2 (6 ''y''1 + 3 ''y''2) / 3, so if the right-hand side is positive, the left-hand side must be positive too. * Otherwise, option 1 is false, since the unique solution of the linear equations is not weakly positive. But in this case, option 2 is true: ** If ''b''2 < 0, then we can take e.g. ''y''1 = 0 and ''y''2 = 1. ** If ''b''1 − 2''b''2 < 0, then, for some number ''B'' > 0, ''b''1 = 2''b''2 − B, so: ''b''1 ''y''1 + ''b''2 ''y''2 = 2 ''b''2 ''y''1 + ''b''2 ''y''2 − ''B'' ''y''1 = ''b''2 (6 ''y''1 + 3 ''y''2) / 3 − ''B'' ''y''1. Thus we can take, for example, ''y''1 = 1, ''y''2 = −2.


Geometric interpretation

Consider the closed convex cone C(\mathbf) spanned by the columns of \mathbf; that is, : C(\mathbf) = \. Observe that C(\mathbf) is the set of the vectors \mathbf for which the first assertion in the statement of Farkas' lemma holds. On the other hand, the vector \mathbf in the second assertion is orthogonal to a hyperplane that separates \mathbf and C(\mathbf). The lemma follows from the observation that \mathbf belongs to C(\mathbf) if and only if there is no hyperplane that separates it from C(\mathbf). More precisely, let \mathbf_1, \dots, \mathbf_n \in \mathbb^ denote the columns of \mathbf. In terms of these vectors, Farkas' lemma states that exactly one of the following two statements is true: # There exist non-negative coefficients x_1, \dots, x_n \in \mathbb such that \mathbf =x _1 \mathbf_1 + \dots + x_n \mathbf_n. # There exists a vector \mathbf \in \mathbb^m such that \mathbf_i^ \mathbf \geq 0 for i = 1, \dots, n, and \mathbf^ \mathbf < 0. The sums x_1 \mathbf_1 + \dots + x_n \mathbf_n with nonnegative coefficients x_1, \dots, x_n form the cone spanned by the columns of \mathbf. Therefore, the first statement tells that \mathbf belongs to C(\mathbf). The second statement tells that there exists a vector \mathbf such that the angle of \mathbf with the vectors \mathbf_i is at most 90°, while the angle of \mathbf with the vector \mathbf is more than 90°. The hyperplane normal to this vector has the vectors \mathbf_i on one side and the vector \mathbf on the other side. Hence, this hyperplane separates the cone spanned by \mathbf_1, \dots, \mathbf_n from the vector \mathbf. For example, let ''n'', ''m'' = 2, ''a''1 = (1, 0)T, and ''a''2 = (1, 1)T. The convex cone spanned by ''a''1 and ''a''2 can be seen as a wedge-shaped slice of the first quadrant in the ''xy'' plane. Now, suppose ''b'' = (0, 1). Certainly, ''b'' is not in the convex cone ''a''1''x''1 + ''a''2''x''2. Hence, there must be a separating hyperplane. Let ''y'' = (1, −1)T. We can see that ''a''1 · ''y'' = 1, ''a''2 · ''y'' = 0, and ''b'' · ''y'' = −1. Hence, the hyperplane with normal ''y'' indeed separates the convex cone ''a''1''x''1 + ''a''2''x''2 from ''b''.


Logic interpretation

A particularly suggestive and easy-to-remember version is the following: if a set of linear inequalities has no solution, then a contradiction can be produced from it by linear combination with nonnegative coefficients. In formulas: if Axb is unsolvable then y^ A = 0, y^ b = -1, y0 has a solution.. Note that y^ A is a combination of the left-hand sides, y^ b a combination of the right-hand side of the inequalities. Since the positive combination produces a zero vector on the left and a −1 on the right, the contradiction is apparent. Thus, Farkas' lemma can be viewed as a theorem of logical completeness: Axb is a set of "axioms", the linear combinations are the "derivation rules", and the lemma says that, if the set of axioms is inconsistent, then it can be refuted using the derivation rules. Pages 81–104.


Variants

The Farkas Lemma has several variants with different sign constraints (the first one is the original version): * Either the system \mathbf = \mathbf has a solution with \mathbf \geq 0 , or the system \mathbf^\mathbf \geq 0 has a solution with \mathbf^ \mathbf < 0. * Either the system \mathbf \leq \mathbf has a solution with \mathbf \geq 0 , or the system \mathbf^\mathbf \geq 0 has a solution with \mathbf^ \mathbf < 0 and \mathbf \geq 0. * Either the system \mathbf \leq \mathbf has a solution with \mathbf \in \mathbb^n , or the system \mathbf^\mathbf = 0 has a solution with \mathbf^ \mathbf < 0 and \mathbf \geq 0. * Either the system \mathbf = \mathbf has a solution with \mathbf \in \mathbb^n , or the system \mathbf^\mathbf = 0 has a solution with \mathbf^ \mathbf \neq 0. The latter variant is mentioned for completeness; it is not actually a "Farkas lemma" since it contains only equalities. Its proof is an exercise in linear algebra.


Generalizations

Generalized Farkas' lemma can be interpreted geometrically as follows: either a vector is in a given closed convex cone, or there exists a hyperplane separating the vector from the cone; there are no other possibilities. The closedness condition is necessary, see Separation theorem I in
Hyperplane separation theorem In geometry, the hyperplane separation theorem is a theorem about disjoint convex sets in ''n''-dimensional Euclidean space. There are several rather similar versions. In one version of the theorem, if both these sets are closed and at least one ...
. For original Farkas' lemma, \mathbf is the nonnegative orthant \mathbb_+^n, hence the closedness condition holds automatically. Indeed, for polyhedral convex cone, i.e., there exists a \mathbf \in \mathbb^ such that \mathbf = \ , the closedness condition holds automatically. In
convex optimization Convex optimization is a subfield of mathematical optimization that studies the problem of minimizing convex functions over convex sets (or, equivalently, maximizing concave functions over convex sets). Many classes of convex optimization pr ...
, various kinds of constraint qualification, e.g.
Slater's condition In mathematics, Slater's condition (or Slater condition) is a sufficient condition for strong duality to hold for a convex optimization problem, named after Morton L. Slater. Informally, Slater's condition states that the feasible region must ...
, are responsible for closedness of the underlying convex cone C(\mathbf). By setting \mathbf = \mathbb^n and \mathbf^* = \ in generalized Farkas' lemma, we obtain the following corollary about the solvability for a finite system of linear equalities:


Further implications

Farkas's lemma can be varied to many further theorems of alternative by simple modifications, such as Gordan's theorem: Either Ax < 0 has a solution ''x'', or A^ y = 0 has a nonzero solution ''y'' with ''y'' ≥ 0. Common applications of Farkas' lemma include proving the strong duality theorem associated with linear programming and the Karush–Kuhn–Tucker conditions. An extension of Farkas' lemma can be used to analyze the strong duality conditions for and construct the dual of a semidefinite program. It is sufficient to prove the existence of the Karush–Kuhn–Tucker conditions using the Fredholm alternative but for the condition to be necessary, one must apply von Neumann's
minimax theorem In the mathematical area of game theory, a minimax theorem is a theorem providing conditions that guarantee that the max–min inequality is also an equality. The first theorem in this sense is von Neumann's minimax theorem from 1928, which was c ...
to show the equations derived by Cauchy are not violated.


See also

*
Dual linear program The dual of a given linear program (LP) is another LP that is derived from the original (the primal) LP in the following schematic way: * Each variable in the primal LP becomes a constraint in the dual LP; * Each constraint in the primal LP becomes ...
*
Fourier–Motzkin elimination Fourier–Motzkin elimination, also known as the FME method, is a mathematical algorithm for eliminating variables from a system of linear inequalities. It can output real solutions. The algorithm is named after Joseph Fourier who proposed the m ...
– can be used to prove Farkas' lemma.


Notes


Further reading

* * {{citation, first = R. T. , last=Rockafellar, author-link=R. T. Rockafellar, title=Convex Analysis, publisher= Princeton University Press, year=1979 , page=200 , mode=cs1 * Kutateladze S.S.br>The Farkas lemma revisited.
''Siberian Mathematical Journal'', 2010, Vol. 51, No. 1, 78–87. Lemmas in linear algebra Convex analysis Linear programming